Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Showing 1 - 5 of 5 matches in All Departments
This book grew out of a NATO Advanced Study Institute summer school that was held in Antalya, TUrkey from 26 May to 6 June 1997. The purpose of the summer school was to expose recent advances in the formal verification of systems composed of both logical and continuous time components. The course was structured in two parts. The first part covered theorem-proving, system automaton models, logics, tools, and complexity of verification. The second part covered modeling and verification of hybrid systems, i. e. , systems composed of a discrete event part and a continuous time part that interact with each other in novel ways. Along with advances in microelectronics, methods to design and build logical systems have grown progressively complex. One way to tackle the problem of ensuring the error-free operation of digital or hybrid systems is through the use of formal techniques. The exercise of comparing the formal specification of a logical system namely, what it is supposed to do to its formal operational description-what it actually does!-in an automated or semi-automated manner is called verification. Verification can be performed in an after-the-fact manner, meaning that after a system is already designed, its specification and operational description are regenerated or modified, if necessary, to match the verification tool at hand and the consistency check is carried out.
This volume contains the proceedings of the second workshop on Computer Aided Verification, held at DIMACS, Rutgers University, June 18-21, 1990. Itfeatures theoretical results that lead to new or more powerful verification methods. Among these are advances in the use of binary decision diagrams, dense time, reductions based upon partial order representations and proof-checking in controller verification. The motivation for holding a workshop on computer aided verification was to bring together work on effective algorithms or methodologies for formal verification - as distinguished, say, from attributes of logics or formal languages. The considerable interest generated by the first workshop, held in Grenoble, June 1989 (see LNCS 407), prompted this second meeting. The general focus of this volume is on the problem of making formal verification feasible for various models of computation. Specific emphasis is on models associated with distributed programs, protocols, and digital circuits. The general test of algorithm feasibility is to embed it into a verification tool, and exercise that tool on realistic examples: the workshop included sessionsfor the demonstration of new verification tools.
This book is for people who want to know what to do with the money they save: so that it's there when they need it - to buy a home, pay for college, etc. - but also grows enough so they don't outlive it. The investment industry is fixated on the importance of maintaining a "balance" of stocks and bonds, shifting to more bonds as one ages. This book challenges this belief by arguing that what's actually important is to have just enough bonds and cash to support spending needs from a stable source, and to replenish these through the sale of stocks at propitious times when the stock market is not depressed. It features simple mathematical calculations, an explanation of basic financial objects like stocks, bonds, ladders, CDs, ETFs, or annuities, a discussion of how to evaluate financial risk, examinations of insurance, fraud deterrence, dollar cost averaging, benefits of a mortgage, risks of a pension, and general advice about healthcare. Although the book is written to be accessible to those with little or no prior knowledge of finance, the studies and conclusions presented here benefit a multitude of financial investors.
Formal verification increasingly has become recognized as an answer to the problem of how to create ever more complex control systems, which nonetheless are required to behave reliably. To be acceptable in an industrial setting, formal verification must be highly algorithmic; to cope with design complexity, it must support a top-down design methodology that leads from an abstract design to its detailed implementation. That combination of requirements points directly to the widely recognized solution of automata-theoretic verification, on account of its expressiveness, computational complexity, and perhaps general utility as well. This book develops the theory of automata-theoretic verification from its foundations, with a focus on algorithms and heuristics to reduce the computational complexity of analysis. It is suitable as a text for a one-or two-semester graduate course, and is recommended reading for anyone planning to use a verification tool, such as COSPAN or SMV. An extensive bibliography that points to the most recent sources, and extensive discussions of methodology and comparisons with other techniques, make this a useful resource for research or verification tool development, as well. Originally published in 1995. The Princeton Legacy Library uses the latest print-on-demand technology to again make available previously out-of-print books from the distinguished backlist of Princeton University Press. These paperback editions preserve the original texts of these important books while presenting them in durable paperback editions. The goal of the Princeton Legacy Library is to vastly increase access to the rich scholarly heritage found in the thousands of books published by Princeton University Press since its founding in 1905.
Formal verification increasingly has become recognized as an answer to the problem of how to create ever more complex control systems, which nonetheless are required to behave reliably. To be acceptable in an industrial setting, formal verification must be highly algorithmic; to cope with design complexity, it must support a top-down design methodology that leads from an abstract design to its detailed implementation. That combination of requirements points directly to the widely recognized solution of automata-theoretic verification, on account of its expressiveness, computational complexity, and perhaps general utility as well. This book develops the theory of automata-theoretic verification from its foundations, with a focus on algorithms and heuristics to reduce the computational complexity of analysis. It is suitable as a text for a one-or two-semester graduate course, and is recommended reading for anyone planning to use a verification tool, such as COSPAN or SMV. An extensive bibliography that points to the most recent sources, and extensive discussions of methodology and comparisons with other techniques, make this a useful resource for research or verification tool development, as well. Originally published in 1995. The Princeton Legacy Library uses the latest print-on-demand technology to again make available previously out-of-print books from the distinguished backlist of Princeton University Press. These editions preserve the original texts of these important books while presenting them in durable paperback and hardcover editions. The goal of the Princeton Legacy Library is to vastly increase access to the rich scholarly heritage found in the thousands of books published by Princeton University Press since its founding in 1905.
|
You may like...
|